54 research outputs found

    Ecological active vision: four bio-inspired principles to integrate bottom-up and adaptive top-down attention tested with a simple camera-arm robot

    Get PDF
    Vision gives primates a wealth of information useful to manipulate the environment, but at the same time it can easily overwhelm their computational resources. Active vision is a key solution found by nature to solve this problem: a limited fovea actively displaced in space to collect only relevant information. Here we highlight that in ecological conditions this solution encounters four problems: 1) the agent needs to learn where to look based on its goals; 2) manipulation causes learning feedback in areas of space possibly outside the attention focus; 3) good visual actions are needed to guide manipulation actions, but only these can generate learning feedback; and 4) a limited fovea causes aliasing problems. We then propose a computational architecture ("BITPIC") to overcome the four problems, integrating four bioinspired key ingredients: 1) reinforcement-learning fovea-based top-down attention; 2) a strong vision-manipulation coupling; 3) bottom-up periphery-based attention; and 4) a novel action-oriented memory. The system is tested with a simple simulated camera-arm robot solving a class of search-and-reach tasks involving color-blob "objects." The results show that the architecture solves the problems, and hence the tasks, very ef?ciently, and highlight how the architecture principles can contribute to a full exploitation of the advantages of active vision in ecological conditions

    Changing perspective on perception physiology: Can you really see what is happening?

    Get PDF
    Perception is a complex, neural mechanism that requires organization and interpretation of input meaning and it has been a key topic in medicine, neuroscience and philosophy for centuries. Gestalt psychology proposed that the underlying mechanism is a constructive process that depends on both input of stimuli and the sensory-motor state of the agent. The Bayesian Brain hypothesis reframed it as probabilistic inference of previous beliefs, which are revised to accommodate new information. The Predictive Coding Theory proposes that this process is implemented through a top-down cascade of cortical predictions of lower level input and the concurrent propagation of a bottom-up prediction error aimed at revising higher level expectations. The „Active Inference‟ theory explains both perception and action, generalising the prediction error minimisation process. In this focused-review we provide a historical overview of the topic and an intuitive approach to the new computational models

    From Psychological Intention Recognition Theories to Adaptive Theory of Mind for Robots

    Get PDF
    Progress in robots' application to everyday scenarios has increased the interest in human-robot interaction (HRI) research. However, robots' limited social skills are associated with decreased humans' positive attitude during HRI. Here, we put forward the idea of developing adaptive Theory of Mind (ToM) model-based systems for social robotics, able to deal with new situations and interact with different users in new tasks. Therefore, we grouped current research from developmental psychology debating the computational processes underlying ToM for HRI strategy development. Defining a model describing adaptive ToM processes may in fact aid the development of adaptive robotic architectures for more flexible and successful HRI. Finally, we hope with this report to both further promote the cross-talk between the fields of developmental psychology and robotics and inspire future investigations in this direction

    How can bottom-up information shape learning of top-down attention-control skills?

    Get PDF
    How does bottom-up information affect the development of top-down attentional control skills during the learning of visuomotor tasks? Why is the eye fovea so small? Strong evidence supports the idea that in humans foveation is mainly guided by task-specific skills, but how these are learned is still an important open problem. We designed and implemented a simulated neural eye-arm coordination model to study the development of attention control in a search-and-reach task involving simple coloured stimuli. The model is endowed with a hard-wired bottom-up attention saliency map and a top-down attention component which acquires task-specific knowledge on potential gaze targets and their spatial relations. This architecture achieves high performance very fast. To explain this result, we argue that: (a) the interaction between bottom-up and top-down mechanisms supports the development of task-specific attention control skills by allowing an efficient exploration of potentially useful gaze targets; (b) bottom-up mechanisms boast the exploitation of the initial limited task-specific knowledge by actively selecting areas where it can be suitably applied; (c) bottom-up processes shape objects representation, their value, and their roles (these can change during learning, e.g. distractors can become useful attentional cues); (d) increasing the size of the fovea alleviates perceptual aliasing, but at the same time increases input processing costs and the number of trials required to learn. Overall, the results indicate that bottom-up attention mechanisms can play a relevant role in attention control, especially during the acquisition of new task-specific skills, but also during task performance

    Learning to look in different environments: an active-vision model which learns and readapts visual routines

    Get PDF
    One of the main claims of the active vision framework is that finding data on the basis of task requirements is more efficient than reconstructing the whole scene by performing a complete visual scan. To be successful, this approach requires that agents learn visual routines to direct overt attention to locations with the information needed to accomplish the task. In ecological conditions, learning such visual routines is difficult due to the partial observability of the world, the changes in the environment, and the fact that learning signals might be indirect. This paper uses a reinforcement-learning actor-critic model to study how visual routines can be formed, and then adapted when the environment changes, in a system endowed with a controllable gaze and reaching capabilities. The tests of the model show that: (a) the autonomously-developed visual routines are strongly dependent on the task and the statistical properties of the environment; (b) when the statistics of the environment change, the performance of the system remains rather stable thanks to the re-use of previously discovered visual routines while the visual exploration policy remains for long time sub-optimal. We conclude that the model has a robust behaviour but the acquisition of an optimal visual exploration policy is particularly hard given its complex dependence on statistical properties of the environment, showing another of the difficulties that adaptive active vision agents must face
    • …
    corecore